88 research outputs found
Jointly Multiple Events Extraction via Attention-based Graph Information Aggregation
Event extraction is of practical utility in natural language processing. In
the real world, it is a common phenomenon that multiple events existing in the
same sentence, where extracting them are more difficult than extracting a
single event. Previous works on modeling the associations between events by
sequential modeling methods suffer a lot from the low efficiency in capturing
very long-range dependencies. In this paper, we propose a novel Jointly
Multiple Events Extraction (JMEE) framework to jointly extract multiple event
triggers and arguments by introducing syntactic shortcut arcs to enhance
information flow and attention-based graph convolution networks to model graph
information. The experiment results demonstrate that our proposed framework
achieves competitive results compared with state-of-the-art methods.Comment: accepted by EMNLP 201
S2F-NER: Exploring Sequence-to-Forest Generation for Complex Entity Recognition
Named Entity Recognition (NER) remains challenging due to the complex
entities, like nested, overlapping, and discontinuous entities. Existing
approaches, such as sequence-to-sequence (Seq2Seq) generation and span-based
classification, have shown impressive performance on various NER subtasks, but
they are difficult to scale to datasets with longer input text because of
either exposure bias issue or inefficient computation. In this paper, we
propose a novel Sequence-to-Forest generation paradigm, S2F-NER, which can
directly extract entities in sentence via a Forest decoder that decode multiple
entities in parallel rather than sequentially. Specifically, our model generate
each path of each tree in forest autoregressively, where the maximum depth of
each tree is three (which is the shortest feasible length for complex NER and
is far smaller than the decoding length of Seq2Seq). Based on this novel
paradigm, our model can elegantly mitigates the exposure bias problem and keep
the simplicity of Seq2Seq. Experimental results show that our model
significantly outperforms the baselines on three discontinuous NER datasets and
on two nested NER datasets, especially for discontinuous entity recognition
Open Domain Event Extraction Using Neural Latent Variable Models
We consider open domain event extraction, the task of extracting unconstraint
types of events from news clusters. A novel latent variable neural model is
constructed, which is scalable to very large corpus. A dataset is collected and
manually annotated, with task-specific evaluation metrics being designed.
Results show that the proposed unsupervised model gives better performance
compared to the state-of-the-art method for event schema induction.Comment: accepted by ACL 201
- …